Sparse Reduced-Rank Regression for Simultaneous Dimension Reduction and Variable Selection in Multivariate Regression
نویسندگان
چکیده
The reduced-rank regression is an effective method to predict multiple response variables from the same set of predictor variables, because it can reduce the number of model parameters as well as take advantage of interrelations between the response variables and therefore improve predictive accuracy. We propose to add a new feature to the reduced-rank regression that allows selection of relevant variables using sparsity inducing penalties. By treating each row of the matrix of the regression coefficients as a group, we propose a group-lasso type penalty and show that this penalty satisfies certain desirable invariance property. We develop two numerical algorithms to solve the penalized regression problem and establish the asymptotic consistency of the proposed method. In particular, the manifold structure of the reduced-rank regression coefficient matrix is respected and carefully studied in our theoretical analysis. In a simulation study and real data analysis, the new method is compared with several existing variable selection methods for multivariate regression and exhibits competitive performance in prediction and variable selection.
منابع مشابه
Bayesian sparse reduced rank multivariate regression
Many modern statistical problems can be cast in the framework of multivariate regression, where the main task is to make statistical inference for a possibly sparse and low-rank coefficient matrix. The low-rank structure in the coefficient matrix is of intrinsic multivariate nature, which, when combined with sparsity, can further lift dimension reduction, conduct variable selection, and facilit...
متن کاملSparse Reduced Rank Regression With Nonconvex Regularization
In this paper, the estimation problem for sparse reduced rank regression (SRRR) model is considered. The SRRR model is widely used for dimension reduction and variable selection with applications in signal processing, econometrics, etc. The problem is formulated to minimize the least squares loss with a sparsity-inducing penalty considering an orthogonality constraint. Convex sparsity-inducing ...
متن کاملSparse partial least squares regression for simultaneous dimension reduction and variable selection
Partial least squares regression has been an alternative to ordinary least squares for handling multicollinearity in several areas of scientific research since the 1960s. It has recently gained much attention in the analysis of high dimensional genomic data. We show that known asymptotic consistency of the partial least squares estimator for a univariate response does not hold with the very lar...
متن کاملScalable Reduced-rank and Sparse Regression
Reduced-rank regression, i.e., multi-task regression subject to a low-rank constraint, is an effective approach to reduce the number of observations required for estimation consistency. However, it is still possible for the estimated singular vectors to be inconsistent in high dimensions as the number of predictors go to infinity with a faster rate than the number of available observations. Spa...
متن کاملExpression quantitative trait loci mapping with multivariate sparse partial least squares regression.
Expression quantitative trait loci (eQTL) mapping concerns finding genomic variation to elucidate variation of expression traits. This problem poses significant challenges due to high dimensionality of both the gene expression and the genomic marker data. We propose a multivariate response regression approach with simultaneous variable selection and dimension reduction for the eQTL mapping prob...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2012